Targeted cross-validation

نویسندگان

چکیده

In many applications, we have access to the complete dataset but are only interested in prediction of a particular region predictor variables. A standard approach is find globally best modeling method from set candidate methods. However, it perhaps rare reality that one uniformly better than others. natural for this scenario apply weighted L2 loss performance assessment reflect region-specific interest. We propose targeted cross-validation (TCV) select models or procedures based on general loss. show TCV consistent selecting performing under Experimental studies used demonstrate use and its potential advantage over global CV using local data region. Previous investigations relied condition when sample size large enough, ranking two candidates stays same. applications with setup changing data-generating processes highly adaptive methods, relative methods not static as varies. Even fixed process, possible switches infinitely times. work, broaden concept selection consistency by allowing switch varies, then establish TCV. This flexible framework can be applied high-dimensional complex machine learning scenarios where performances dynamic.

برای دانلود باید عضویت طلایی داشته باشید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Customer Validation in Cross-Dock

Considering the importance of validation of customers in the cross-dock and since this is one of the problems of implementing cross-dock system in Iran, this study attempted to extract customer validation criteria. The purpose of the research is to eliminate the distrust of distributors in receiving the funds of the sent items and the statistical sample of this research is the experts of the sy...

متن کامل

Cross-Validation Without Doing Cross-Validation in Genome-Enabled Prediction

Cross-validation of methods is an essential component of genome-enabled prediction of complex traits. We develop formulae for computing the predictions that would be obtained when one or several cases are removed in the training process, to become members of testing sets, but by running the model using all observations only once. Prediction methods to which the developments apply include least ...

متن کامل

Insights into Cross-validation

Cross-validation is one of the most widely used techniques, in estimating the Generalization Error of classification algorithms. Though several empirical studies have been conducted, to study the behavior of this method in the past, none of them clearly elucidate the reasons behind the observed behavior. In this paper we study the behavior of the moments (i.e. expected value and variance) of th...

متن کامل

Cross-Lingual Answer Validation

We describe three language-independent methods for the task of answer validation. All methods are based on a scoring mechanism that reflects the degree of similarity between the question-answer pairs and the supporting text. We evaluate the proposed methods when using various string similarity metrics, such as exact matching, Levenshtein, Jaro and Jaro-Winkler. In addition to this baseline appr...

متن کامل

Cross-Validation with LULOO

| The leave-one-out cross-validation scheme for generalization assessment of neu-ral network models is computationally expensive due to replicated training sessions. Linear unlearning of examples has recently been suggested as an approach to approximative cross-validation. Here we brieey review the linear unlearning scheme, dubbed LULOO, and we illustrate it on a system identiication example. F...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: Bernoulli

سال: 2023

ISSN: ['1573-9759', '1350-7265']

DOI: https://doi.org/10.3150/22-bej1461